Goto

Collaborating Authors

 lder space




Adaptation to Misspecified Kernel Regularity in Kernelised Bandits

Liu, Yusha, Singh, Aarti

arXiv.org Artificial Intelligence

In continuum-armed bandit problems where the underlying function resides in a reproducing kernel Hilbert space (RKHS), namely, the kernelised bandit problems, an important open problem remains of how well learning algorithms can adapt if the regularity of the associated kernel function is unknown. In this work, we study adaptivity to the regularity of translation-invariant kernels, which is characterized by the decay rate of the Fourier transformation of the kernel, in the bandit setting. We derive an adaptivity lower bound, proving that it is impossible to simultaneously achieve optimal cumulative regret in a pair of RKHSs with different regularities. To verify the tightness of this lower bound, we show that an existing bandit model selection algorithm applied with minimax non-adaptive kernelised bandit algorithms matches the lower bound in dependence of $T$, the total number of steps, except for log factors. By filling in the regret bounds for adaptivity between RKHSs, we connect the statistical difficulty for adaptivity in continuum-armed bandits in three fundamental types of function spaces: RKHS, Sobolev space, and H\"older space.


Asymptotic Properties for Bayesian Neural Network in Besov Space

Lee, Kyeongwon, Lee, Jaeyong

arXiv.org Artificial Intelligence

Neural networks have shown great predictive power when dealing with various unstructured data such as images and natural languages. The Bayesian neural network captures the uncertainty of prediction by putting a prior distribution for the parameter of the model and computing the posterior distribution. In this paper, we show that the Bayesian neural network using spike-and-slab prior has consistency with nearly minimax convergence rate when the true regression function is in the Besov space. Even when the smoothness of the regression function is unknown the same posterior convergence rate holds and thus the spike-and-slab prior is adaptive to the smoothness of the regression function. We also consider the shrinkage prior, which is more feasible than other priors, and show that it has the same convergence rate. In other words, we propose a practical Bayesian neural network with guaranteed asymptotic properties.


Smooth Bandit Optimization: Generalization to H\"older Space

Liu, Yusha, Wang, Yining, Singh, Aarti

arXiv.org Machine Learning

We consider bandit optimization of a smooth reward function, where the goal is cumulative regret minimization. This problem has been studied for $\alpha$-H\"older continuous (including Lipschitz) functions with $0<\alpha\leq 1$. Our main result is in generalization of the reward function to H\"older space with exponent $\alpha>1$ to bridge the gap between Lipschitz bandits and infinitely-differentiable models such as linear bandits. For H\"older continuous functions, approaches based on random sampling in bins of a discretized domain suffices as optimal. In contrast, we propose a class of two-layer algorithms that deploy misspecified linear/polynomial bandit algorithms in bins. We demonstrate that the proposed algorithm can exploit higher-order smoothness of the function by deriving a regret upper bound of $\tilde{O}(T^\frac{d+\alpha}{d+2\alpha})$ for when $\alpha>1$, which matches existing lower bound. We also study adaptation to unknown function smoothness over a continuous scale of H\"older spaces indexed by $\alpha$, with a bandit model selection approach applied with our proposed two-layer algorithms. We show that it achieves regret rate that matches the existing lower bound for adaptation within the $\alpha\leq 1$ subset.


Continuum-Armed Bandits: A Function Space Perspective

Singh, Shashank

arXiv.org Machine Learning

Continuum-armed bandits (a.k.a., black-box or $0^{th}$-order optimization) involves optimizing an unknown objective function given an oracle that evaluates the function at a query point, with the goal of using as few query points as possible. In the most well-studied case, the objective function is assumed to be Lipschitz continuous and minimax rates of simple and cumulative regrets are known in both noiseless and noisy settings. This paper studies continuum-armed bandits under more general smoothness conditions, namely Besov smoothness conditions, on the objective function. In both noiseless and noisy conditions, we derive minimax rates under simple and cumulative regrets. Our results show that minimax rates over objective functions in a Besov space are identical to minimax rates over objective functions in the smallest H\"older space into which the Besov space embeds.


Rate-Optimal Detection of Very Short Signal Segments

Cai, T. Tony, Yuan, Ming

arXiv.org Machine Learning

Detection of very short signal segments arise in a wide range of applications in many fields including engineering, genomics, and material science. For example, copy number variations (CNVs) play a significant role in the genetics of complex disease. Therefore the detection of CNVs due to duplication and deletion of a segment of DNA sequences is an important problem in genomics. In contrast to single-nucleotide polymorphisms which affects only one single nucleotide base, each CNV corresponds to a short segment of the genome, typically around 1000 nucleotide bases, that has been altered (see, e.g., Stankiewicz and Lupski, 2010). Although the length of these CNVs is much smaller than that of the whole genome, recognizing and accounting for such segment structure are critical in effective detection of CNVs (see, e.g., Jeng, Cai and Li, 2010).